An Efficient Sparse Metric Learning in High-Dimensional Space via `1-Penalized Log-Determinant Regularization
نویسندگان
چکیده
This paper proposes an efficient sparse metric learning algorithm in high dimensional space via an `1-penalized log-determinant regularization. Compare to the most existing distance metric learning algorithms, the proposed algorithm exploits the sparsity nature underlying the intrinsic high dimensional feature space. This sparsity prior of learning distance metric serves to regularize the complexity of the distance model especially in the “less example number p and high dimension d” setting. Theoretically, by analogy to the covariance estimation problem, we find the proposed distance learning algorithm has a consistent result at rate O (√( m log d )/ n ) to the target distance matrix with at most m nonzeros per row. Moreover, from the implementation perspective, this `1-penalized logdeterminant formulation can be efficiently optimized in a block coordinate descent fashion which is much faster than the standard semi-definite programming which has been widely adopted in many other advanced distance learning algorithms. We compare this algorithm with other state-of-the-art ones on various datasets and competitive results are obtained.
منابع مشابه
Vector ℓ0 latent-space principal component analysis
Principal component analysis (PCA) is a widely used signal processing technique. Instead of performing PCA in the data space, we consider the problem of sparse PCA in a potentially higher dimensional latent space. To do so, we zero-out groups of variables using vector `0 regularization. The estimation is based on the maximization of the penalized log-likelihood, for which we develop an efficien...
متن کاملPattern alternating maximization algorithm for missing data in high-dimensional problems
We propose a novel and efficient algorithm for maximizing the observed log-likelihood of a multivariate normal data matrix with missing values. We show that our procedure, based on iteratively regressing the missing on the observed variables, generalizes the standard EM algorithm by alternating between different complete data spaces and performing the E-Step incrementally. In this non-standard ...
متن کاملOn the Conditions of Sparse Parameter Estimation via Log-Sum Penalty Regularization
For high-dimensional sparse parameter estimation problems, Log-Sum Penalty (LSP) regularization effectively reduces the sampling sizes in practice. However, it still lacks theoretical analysis to support the experience from previous empirical study. The analysis of this article shows that, like `0-regularization, O(s) sampling size is enough for proper LSP, where s is the non-zero components of...
متن کاملMammalian Eye Gene Expression Using Support Vector Regression to Evaluate a Strategy for Detecting Human Eye Disease
Background and purpose: Machine learning is a class of modern and strong tools that can solve many important problems that nowadays humans may be faced with. Support vector regression (SVR) is a way to build a regression model which is an incredible member of the machine learning family. SVR has been proven to be an effective tool in real-value function estimation. As a supervised-learning appr...
متن کاملSparse Distance Weighted Discrimination
Distance weighted discrimination (DWD) was originally proposed to handle the data piling issue in the support vector machine. In this paper, we consider the sparse penalized DWD for high-dimensional classification. The state-of-the-art algorithm for solving the standard DWD is based on second-order cone programming, however such an algorithm does not work well for the sparse penalized DWD with ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2009